Uncertainty-Aware Self-Training for Low-Resource Neural Sequence Labeling

نویسندگان

چکیده

Neural sequence labeling (NSL) aims at assigning labels for input language tokens, which covers a broad range of applications, such as named entity recognition (NER) and slot filling, etc. However, the satisfying results achieved by traditional supervised-based approaches heavily depend on large amounts human annotation data, may not be feasible in real-world scenarios due to data privacy computation efficiency issues. This paper presents SeqUST, novel uncertain-aware self-training framework NSL address labeled scarcity issue effectively utilize unlabeled data. Specifically, we incorporate Monte Carlo (MC) dropout Bayesian neural network (BNN) perform uncertainty estimation token level then select reliable tokens from based model confidence certainty. A well-designed masked task with noise-robust loss supports robust training, suppress problem noisy pseudo labels. In addition, develop Gaussian-based consistency regularization technique further improve robustness Gaussian-distributed perturbed representations. alleviates over-fitting dilemma originating pseudo-labeled augmented Extensive experiments over six benchmarks demonstrate that our SeqUST improves performance self-training, consistently outperforms strong baselines margin low-resource scenarios.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Empower Sequence Labeling with Task-Aware Neural Language Model

Linguistic sequence labeling is a general modeling approach that encompasses a variety of problems, such as part-ofspeech tagging and named entity recognition. Recent advances in neural networks (NNs) make it possible to build reliable models without handcrafted features. However, in many cases, it is hard to obtain sufficient annotations to train these models. In this study, we develop a novel...

متن کامل

Low-Resource Semantic Role Labeling

We explore the extent to which highresource manual annotations such as treebanks are necessary for the task of semantic role labeling (SRL). We examine how performance changes without syntactic supervision, comparing both joint and pipelined methods to induce latent syntax. This work highlights a new application of unsupervised grammar induction and demonstrates several approaches to SRL in the...

متن کامل

Implicitly-Defined Neural Networks for Sequence Labeling

In this work, we propose a novel, implicitlydefined neural network architecture and describe a method to compute its components. The proposed architecture forgoes the causality assumption previously used to formulate recurrent neural networks and allow the hidden states of the network to coupled together, allowing potential improvement on problems with complex, long-distance dependencies. Initi...

متن کامل

Adapting Self-Training for Semantic Role Labeling

Supervised semantic role labeling (SRL) systems trained on hand-crafted annotated corpora have recently achieved state-of-the-art performance. However, creating such corpora is tedious and costly, with the resulting corpora not sufficiently representative of the language. This paper describes a part of an ongoing work on applying bootstrapping methods to SRL to deal with this problem. Previous ...

متن کامل

Analysis of low-resource acoustic model self-training

Previous work on self-training of acoustic models using unlabeled data reported significant reductions in WER assuming a large phonetic dictionary was available. We now assume only those words from ten hours of speech are initially available. Subsequently, we are then given a large vocabulary and then quantify the value of repeating selftraining with this larger dictionary. This experiment is u...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i11.26603